# English vision model
CLIP ViT H 14 Laion2b S32b B79k
MIT
This is a vision-language model based on the OpenCLIP framework, trained on the LAION-2B English subset, excelling in zero-shot image classification and cross-modal retrieval tasks.
Text-to-Image
C
ModelsLab
132
0
CLIP ViT B 32 Laion2b S34b B79k
MIT
CLIP ViT-B/32 model trained on the LAION-2B dataset based on the OpenCLIP framework, supporting zero-shot image classification and cross-modal retrieval tasks
Text-to-Image
C
rroset
48
0
Featured Recommended AI Models